48

Algorithms for Binary Neural Networks

FIGURE 3.7

Accuracy with different θ and λ.

TABLE 3.1

Classification accuracy (%) on CIFAR-10 with 20-layer U-MCNs and MCNs.

Method

Kernel Stage

Size (MB)

WRNs

U-MCNs

MCNs

MCNs-1

MCNs

16-16-32-64

1.1

92.31

93.69

92.08

92.10

16-32-64-128

4.3

94.88

93.98

93.94

32-64-128-256

17.1

95.50

95.13

95.33

64-64-128-256

17.2

95.75

95.72

95.30

95.34

LBCNN (q=384)

61

92.96

respectively, with similar accuracy (93.98% vs. 92.96%). When LBCNN has several param-

eters (4.3M) similar to those of the MCNs, the test run time of LBCNN becomes 16.2 s,

which is still slower than our MCNs.

Visualization: We visualize MCconv features in Fig. 3.8 across different layers and the

curves of elements in different M-Filters in Fig. 3.11. Similarly to conventional CNNs,

the features of different layers capture rich and hierarchy information in Fig. 3.8. Based

on the reconstructed filters Q corresponding to the M-Filters, we obtain convolutional fea-

tures that appear diverse for different M-Filters. In summary, different MCconv layers and

Input

MCconv 1

MCconv 2

MCconv 3

FIGURE 3.8

Example of output feature maps produced by Q from different layers.